Direct Density-Derivative Estimation and Its Application in KL-Divergence Approximation
نویسندگان
چکیده
Estimation of density derivatives is a versatile tool in statistical data analysis. A naive approach is to first estimate the density and then compute its derivative. However, such a two-step approach does not work well because a good density estimator does not necessarily mean a good densityderivative estimator. In this paper, we give a direct method to approximate the density derivative without estimating the density itself. Our proposed estimator allows analytic and computationally efficient approximation of multi-dimensional high-order density derivatives, with the ability that all hyper-parameters can be chosen objectively by cross-validation. We further show that the proposed density-derivative estimator is useful in improving the accuracy of non-parametric KL-divergence estimation via metric learning. The practical superiority of the proposed method is experimentally demonstrated in change detection and feature selection.
منابع مشابه
Robust Estimation in Linear Regression Model: the Density Power Divergence Approach
The minimum density power divergence method provides a robust estimate in the face of a situation where the dataset includes a number of outlier data. In this study, we introduce and use a robust minimum density power divergence estimator to estimate the parameters of the linear regression model and then with some numerical examples of linear regression model, we show the robustness of this est...
متن کاملPAC-Bayesian Generalization Bound for Density Estimation with Application to Co-clustering
We derive a PAC-Bayesian generalization bound for density estimation. Similar to the PAC-Bayesian generalization bound for classification, the result has the appealingly simple form of a tradeoff between empirical performance and the KL-divergence of the posterior from the prior. Moreover, the PACBayesian generalization bound for classification can be derived as a special case of the bound for ...
متن کاملIndependent component analysis based on marginal density estimation using weighted Parzen windows
This work proposes a novel algorithm for independent component analysis (ICA) based on marginal density estimation. The proposed ICA algorithm aims to search for an effective demixing matrix as well as weighted Parzen window (WPW) representations for marginal densities of independent components so as to express a factorial joint density for high dimensional observations. Following the linear mi...
متن کاملBias Reduction and Metric Learning for Nearest-Neighbor Estimation of Kullback-Leibler Divergence
Asymptotically unbiased nearest-neighbor estimators for KL divergence have recently been proposed and demonstrated in a number of applications. With small sample sizes, however, these nonparametric methods typically suffer from high estimation bias due to the non-local statistics of empirical nearest-neighbor information. In this paper, we show that this non-local bias can be mitigated by chang...
متن کاملTree-structured Gaussian process approximations Supplementary material
subject to q(f |u) = ∏ i q(fi|u) and ∫ dfiq(fi|u) = 1. It is noted that KL(a||b) is the measurement of information “lost” when using b to approximate a. It was argued in [1] that it is appropriate to use this KL divergence as an approximation measure since we are trying to find a sparse representation u and its relationship with f to approximate p by q. The KL divergence above can be expanded a...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2015